Magnum V4 72b FP8 Dynamic
Apache-2.0
A large language model with 72B parameters fine - tuned based on Qwen2.5 - 72B - Instruct. It uses dynamic FP8 quantization technology to optimize inference efficiency and aims to reproduce the prose quality of Claude 3.
Large Language Model
Transformers English